383 research outputs found

    Learning Social Image Embedding with Deep Multimodal Attention Networks

    Full text link
    Learning social media data embedding by deep models has attracted extensive research interest as well as boomed a lot of applications, such as link prediction, classification, and cross-modal search. However, for social images which contain both link information and multimodal contents (e.g., text description, and visual content), simply employing the embedding learnt from network structure or data content results in sub-optimal social image representation. In this paper, we propose a novel social image embedding approach called Deep Multimodal Attention Networks (DMAN), which employs a deep model to jointly embed multimodal contents and link information. Specifically, to effectively capture the correlations between multimodal contents, we propose a multimodal attention network to encode the fine-granularity relation between image regions and textual words. To leverage the network structure for embedding learning, a novel Siamese-Triplet neural network is proposed to model the links among images. With the joint deep model, the learnt embedding can capture both the multimodal contents and the nonlinear network information. Extensive experiments are conducted to investigate the effectiveness of our approach in the applications of multi-label classification and cross-modal search. Compared to state-of-the-art image embeddings, our proposed DMAN achieves significant improvement in the tasks of multi-label classification and cross-modal search

    The Pragmatic Function of Vague Language in Business Letters Guided by the Relevance Theory

    Get PDF
    Trade between China and foreign countries is increasingly frequent. As the medium of communication, business letters are playing a more and more important role in trading. It is easy to find that vague language is widely used in business letters. The successful use of vague language requires the recipient to be able to infer the writer’s intention correctly. This paper takes business letters collected online and offline as corpus, and will analyze the vague language from two aspects, that is, lexical aspect and syntactic aspect. Based on the analysis finally summarize the three main pragmatic functions of vague language in business letters, namely intent expression, self-protection and politeness

    Supervised, semi-supervised, and unsupervised learning of the Domany-Kinzel model

    Full text link
    The Domany Kinzel (DK) model encompasses several types of non-equilibrium phase transitions, depending on the selected parameters. We apply supervised, semi-supervised, and unsupervised learning methods to studying the phase transitions and critical behaviors of the (1 + 1)-dimensional DK model. The supervised and the semi-supervised learning methods permit the estimations of the critical points, the spatial and temporal correlation exponents, concerning labelled and unlabelled DK configurations, respectively. Furthermore, we also predict the critical points by employing principal component analysis (PCA) and autoencoder. The PCA and autoencoder can produce results in good agreement with simulated particle number density

    Deep Residual Shrinkage Networks for EMG-based Gesture Identification

    Full text link
    This work introduces a method for high-accuracy EMG based gesture identification. A newly developed deep learning method, namely, deep residual shrinkage network is applied to perform gesture identification. Based on the feature of EMG signal resulting from gestures, optimizations are made to improve the identification accuracy. Finally, three different algorithms are applied to compare the accuracy of EMG signal recognition with that of DRSN. The result shows that DRSN excel traditional neural networks in terms of EMG recognition accuracy. This paper provides a reliable way to classify EMG signals, as well as exploring possible applications of DRSN

    Study on the Impact of Private Enterprises\u27 Participation in the Mixed Reform of State-owned Enterprises on the Value Preservation and Appreciation of State-owned Assets

    Get PDF
    State-owned listed companies in the commercial category in Shanghai and Shenzhen A-shares from 2013 to 2020 were selected as the research sample. Data related to the shareholding and delegated behavior of private shareholders among the top ten shareholders are utilized. Using a fixed-effects model, we empirically analyze the impact of private companies\u27 participation in the mixed reform of state-owned enterprises on the value preservation and appreciation of state-owned assets at the equity level and the management right level. Explore the moderating effects of internal control and the level of market-oriented development on the above relationships. The study shows that the participation of private enterprises in the mixed reform of SOEs can effectively contribute to the value preservation and appreciation of state-owned assets, both at the equity level and at the management level. Meanwhile, the level of internal control of SOEs has a significant positive moderating effect on the relationship between the participation of private firms in SOEs\u27 mixed reform and the value-added of SOEs\u27 assets; The participation of private firms can effectively compensate for the impact of insufficient market development on the value-added of SOEs\u27 assets

    Efficient Privacy-Preserving Machine Learning with Lightweight Trusted Hardware

    Full text link
    In this paper, we propose a new secure machine learning inference platform assisted by a small dedicated security processor, which will be easier to protect and deploy compared to today's TEEs integrated into high-performance processors. Our platform provides three main advantages over the state-of-the-art: (i) We achieve significant performance improvements compared to state-of-the-art distributed Privacy-Preserving Machine Learning (PPML) protocols, with only a small security processor that is comparable to a discrete security chip such as the Trusted Platform Module (TPM) or on-chip security subsystems in SoCs similar to the Apple enclave processor. In the semi-honest setting with WAN/GPU, our scheme is 4X-63X faster than Falcon (PoPETs'21) and AriaNN (PoPETs'22) and 3.8X-12X more communication efficient. We achieve even higher performance improvements in the malicious setting. (ii) Our platform guarantees security with abort against malicious adversaries under honest majority assumption. (iii) Our technique is not limited by the size of secure memory in a TEE and can support high-capacity modern neural networks like ResNet18 and Transformer. While previous work investigated the use of high-performance TEEs in PPML, this work represents the first to show that even tiny secure hardware with really limited performance can be leveraged to significantly speed-up distributed PPML protocols if the protocol can be carefully designed for lightweight trusted hardware.Comment: IEEE S&P'24 submitte

    JoinGym: An Efficient Query Optimization Environment for Reinforcement Learning

    Full text link
    In this paper, we present \textsc{JoinGym}, an efficient and lightweight query optimization environment for reinforcement learning (RL). Join order selection (JOS) is a classic NP-hard combinatorial optimization problem from database query optimization and can serve as a practical testbed for the generalization capabilities of RL algorithms. We describe how to formulate each of the left-deep and bushy variants of the JOS problem as a Markov Decision Process (MDP), and we provide an implementation adhering to the standard Gymnasium API. We highlight that our implementation \textsc{JoinGym} is completely based on offline traces of all possible joins, which enables RL practitioners to easily and quickly test their methods on a realistic data management problem without needing to setup any systems. Moreover, we also provide all possible join traces on 33003300 novel SQL queries generated from the IMDB dataset. Upon benchmarking popular RL algorithms, we find that at least one method can obtain near-optimal performance on train-set queries but their performance degrades by several orders of magnitude on test-set queries. This gap motivates further research for RL algorithms that generalize well in multi-task combinatorial optimization problems.Comment: We will make all the queries available soo
    • …
    corecore